Talking Face Generation


Talking face generation is the process of generating videos of a person speaking based on an audio recording of their voice.

JoyGen: Audio-Driven 3D Depth-Aware Talking-Face Video Editing

Add code
Jan 03, 2025
Figure 1 for JoyGen: Audio-Driven 3D Depth-Aware Talking-Face Video Editing
Figure 2 for JoyGen: Audio-Driven 3D Depth-Aware Talking-Face Video Editing
Figure 3 for JoyGen: Audio-Driven 3D Depth-Aware Talking-Face Video Editing
Figure 4 for JoyGen: Audio-Driven 3D Depth-Aware Talking-Face Video Editing
Viaarxiv icon

MoEE: Mixture of Emotion Experts for Audio-Driven Portrait Animation

Add code
Jan 03, 2025
Viaarxiv icon

DEGSTalk: Decomposed Per-Embedding Gaussian Fields for Hair-Preserving Talking Face Synthesis

Add code
Dec 28, 2024
Viaarxiv icon

GLCF: A Global-Local Multimodal Coherence Analysis Framework for Talking Face Generation Detection

Add code
Dec 18, 2024
Figure 1 for GLCF: A Global-Local Multimodal Coherence Analysis Framework for Talking Face Generation Detection
Figure 2 for GLCF: A Global-Local Multimodal Coherence Analysis Framework for Talking Face Generation Detection
Figure 3 for GLCF: A Global-Local Multimodal Coherence Analysis Framework for Talking Face Generation Detection
Figure 4 for GLCF: A Global-Local Multimodal Coherence Analysis Framework for Talking Face Generation Detection
Viaarxiv icon

Joint Co-Speech Gesture and Expressive Talking Face Generation using Diffusion with Adapters

Add code
Dec 18, 2024
Figure 1 for Joint Co-Speech Gesture and Expressive Talking Face Generation using Diffusion with Adapters
Figure 2 for Joint Co-Speech Gesture and Expressive Talking Face Generation using Diffusion with Adapters
Figure 3 for Joint Co-Speech Gesture and Expressive Talking Face Generation using Diffusion with Adapters
Figure 4 for Joint Co-Speech Gesture and Expressive Talking Face Generation using Diffusion with Adapters
Viaarxiv icon

PortraitTalk: Towards Customizable One-Shot Audio-to-Talking Face Generation

Add code
Dec 10, 2024
Figure 1 for PortraitTalk: Towards Customizable One-Shot Audio-to-Talking Face Generation
Figure 2 for PortraitTalk: Towards Customizable One-Shot Audio-to-Talking Face Generation
Figure 3 for PortraitTalk: Towards Customizable One-Shot Audio-to-Talking Face Generation
Figure 4 for PortraitTalk: Towards Customizable One-Shot Audio-to-Talking Face Generation
Viaarxiv icon

VQTalker: Towards Multilingual Talking Avatars through Facial Motion Tokenization

Add code
Dec 13, 2024
Viaarxiv icon

GoHD: Gaze-oriented and Highly Disentangled Portrait Animation with Rhythmic Poses and Realistic Expression

Add code
Dec 13, 2024
Figure 1 for GoHD: Gaze-oriented and Highly Disentangled Portrait Animation with Rhythmic Poses and Realistic Expression
Figure 2 for GoHD: Gaze-oriented and Highly Disentangled Portrait Animation with Rhythmic Poses and Realistic Expression
Figure 3 for GoHD: Gaze-oriented and Highly Disentangled Portrait Animation with Rhythmic Poses and Realistic Expression
Figure 4 for GoHD: Gaze-oriented and Highly Disentangled Portrait Animation with Rhythmic Poses and Realistic Expression
Viaarxiv icon

SINGER: Vivid Audio-driven Singing Video Generation with Multi-scale Spectral Diffusion Model

Add code
Dec 04, 2024
Viaarxiv icon

IF-MDM: Implicit Face Motion Diffusion Model for High-Fidelity Realtime Talking Head Generation

Add code
Dec 05, 2024
Figure 1 for IF-MDM: Implicit Face Motion Diffusion Model for High-Fidelity Realtime Talking Head Generation
Figure 2 for IF-MDM: Implicit Face Motion Diffusion Model for High-Fidelity Realtime Talking Head Generation
Figure 3 for IF-MDM: Implicit Face Motion Diffusion Model for High-Fidelity Realtime Talking Head Generation
Figure 4 for IF-MDM: Implicit Face Motion Diffusion Model for High-Fidelity Realtime Talking Head Generation
Viaarxiv icon